
Prompt-based multi-task learning for robust text retrieval

Annotation
The exponential growth of digital information necessitates the development of robust text retrieval methods since most of the methods are domain or task-specific which limits their implementation. In this case multi-task learning is a promising alternative as it helps a model to have more meaningful embeddings; however such cases require usage of task separation methods. Many studies explore multi-task learning to improve generalization but tend to focus on large models. However, in real-world, speech analytics tasks that require searching through hundreds of millions of vectors in real-time, smaller models become more appropriate. This paper presents a novel approach to enhance the robustness of multi-task text retrieval models through the use of prompts. We use contrastive learning to train encoder models both in single-task and multi-task configurations and compare their performances as well as analyze the efficiency of different prompt usage strategies including hard prompts represented by explicit natural language instructions and soft prompts of varying lengths represented by model special tokens. Experiments are conducted by applying prompts to both the query and candidate document as well as to queries only keeping the candidate without prompts to reuse pre-encoded candidates in multi-task retrieval without significant quality loss. The obtained results are compared using R@1, R@5, and MRR metrics which are most applicable for evaluating in-domain and out-of-domain search. Single-task models show better performance on in-domain training data, while multi-task models demonstrate superior performance on out-of-domain data highlighting their increased robustness to domain shifts. Applying prompts to both elements–query and document–yields better performance than applying them solely to the query. Soft prompts are found to be preferable to hard as they better adapt the model to different domains. The findings of this study can be useful for improving text retrieval models, especially in scenarios involving multi-task systems where high adaptability and performance on new data are required. Trainable prompts could be an effective tool for enhancing the flexibility of models in various applications, such as information retrieval and question-answering systems.
Keywords
Постоянный URL
Articles in current issue
- Design and fabrication of collimating ball lensed fiber for the system of optical radiation output from radiophotonic components
- From Triassic to moderinity: Raman spectroscopy for differentiation of fossil resins age by age
- Optimization of geometry of two-dimensional photonic crystal waveguide for telecommunications and sensorics
- Development and investigation of the suppressing additive noises methods in fiber-optic interferometric sensors
- Method for compensating the constant component of noise in the reflectogram of a fiber-optic communication line under conditions of insufficient dynamic range of an optical backscatter reflectometer in the time domain
- Investigation of the method of moving object weight measurement based onquasidistributed fiber Bragg gratings with temperature compensation
- Modern optical methods of non-contact geometric measurements and reconstruction of object 3D surface shape: a review
- Spectral-luminescent properties of silver clusters Ag1–5 in the ion-exchange layer of silicate glass
- Forming a thick layer of ε-Ga2O3 on the GaN sublayer with V-defects at the interface
- A model for ensuring the continuity of the safe functioning of the product quality traceability system in conditions of unstable communication
- Application of Markov chain Monte Carlo and machine learning for identifying active modules in biological graphs
- Surface defect detection with limited data based on SSD detector and Siamese networks
- Sentiment analysis of Arabic tweets using supervised machine learning (in English)
- Russian parametric corpus RuParam
- Comparative analysis of AI-generated and original abstracts of academic articles on philology
- Enhancing Kubernetes security with machine learning: а proactive approach to anomaly detection
- Improving question answering in programming domain with pretrained language model finetuning using structured diverse online forum data
- Specification language for automatа-based objects cooperation
- Aspects of organizing game interactions among asymmetric agents using graph neural networks
- Development and modeling of technological scheme of steam methane reforming with oxy-fuel combustion and carbon capture
- Stability study of hybrid MOS memristor memory using modified particle swarm optimization method
- Analysis of the vulnerability of YOLO neural network models to the Fast Sign Gradient Method attack